Transformers have recently gained attention in the computer vision domain due to their ability to model long-range dependencies. However, the self-attention mechanism, which is the core part of the Transformer model, usually suffers from quadratic computational complexity with respect to the number of tokens. Many architectures attempt to reduce model complexity by limiting the self-attention mechanism to local regions or by redesigning the tokenization process. In this paper, we propose DAE-Former, a novel method that seeks to provide an alternative perspective by efficiently designing the self-attention mechanism. More specifically, we reformulate the self-attention mechanism to capture both spatial and channel relations across the whole feature dimension while staying computationally efficient. Furthermore, we redesign the skip connection path by including the cross-attention module to ensure the feature reusability and enhance the localization power. Our method outperforms state-of-the-art methods on multi-organ cardiac and skin lesion segmentation datasets without requiring pre-training weights. The code is publicly available at https://github.com/mindflow-institue/DAEFormer.
translated by 谷歌翻译
In this paper, we present a novel control architecture for the online adaptation of bipedal locomotion on inclined obstacles. In particular, we introduce a novel, cost-effective, and versatile foot sensor to detect the proximity of the robot's feet to the ground (bump sensor). By employing this sensor, feedback controllers are implemented to reduce the impact forces during the transition of the swing to stance phase or steeping on inclined unseen obstacles. Compared to conventional sensors based on contact reaction force, this sensor detects the distance to the ground or obstacles before the foot touches the obstacle and therefore provides predictive information to anticipate the obstacles. The controller of the proposed bump sensor interacts with another admittance controller to adjust leg length. The walking experiments show successful locomotion on the unseen inclined obstacle without reducing the locomotion speed with a slope angle of 12. Foot position error causes a hard impact with the ground as a consequence of accumulative error caused by links and connections' deflection (which is manufactured by university tools). The proposed framework drastically reduces the feet' impact with the ground.
translated by 谷歌翻译
步行运动计划基于运动的不同组成部分(DCM)和线性倒置模型(LIPM)是可以实现的替代方案之一,以生成在线人类人体机器人步态轨迹。该算法需要调整不同的参数。在此,我们开发了一个框架来获得最佳参数,以实现Real Robot步态的稳定且节能的轨迹。为了找到最佳轨迹,在机器人的每个下肢关节下,代表能耗的四个成本函数,关节速度和应用扭矩的总和,以及基于零矩(ZMP)稳定性标准的成本函数。遗传算法用于框架中,以优化这些成本函数中的每一个。尽管轨迹计划是在简化模型的帮助下完成的,但通过考虑Bullet Physics Engine Simulator中的完整动力学模型和脚部接触模型,可以获得每个成本函数的值。这种优化的结果是,以最有效的方式行走的最稳定性和行走是相互对比的。因此,在另一次尝试中,对ZMP和以三种不同速度的能量成本函数进行了多目标优化。最后,我们比较了使用最佳参数生成的设计轨迹,并将模拟产生的仿真模拟器。
translated by 谷歌翻译
卷积神经网络(CNN)已成为医疗图像分割任务的共识。但是,由于卷积操作的性质,它们在建模长期依赖性和空间相关性时受到限制。尽管最初开发了变压器来解决这个问题,但它们未能捕获低级功能。相比之下,证明本地和全球特征对于密集的预测至关重要,例如在具有挑战性的环境中细分。在本文中,我们提出了一种新型方法,该方法有效地桥接了CNN和用于医学图像分割的变压器。具体而言,我们使用开创性SWIN变压器模块和一个基于CNN的编码器设计两个多尺度特征表示。为了确保从上述两个表示获得的全局和局部特征的精细融合,我们建议在编码器编码器结构的跳过连接中提出一个双层融合(DLF)模块。在各种医学图像分割数据集上进行的广泛实验证明了Hiformer在计算复杂性以及定量和定性结果方面对其他基于CNN的,基于变压器和混合方法的有效性。我们的代码可在以下网址公开获取:https://github.com/amirhossein-kz/hiformer
translated by 谷歌翻译
The field of artificial intelligence (AI), regarded as one of the most enigmatic areas of science, has witnessed exponential growth in the past decade including a remarkably wide array of applications, having already impacted our everyday lives. Advances in computing power and the design of sophisticated AI algorithms have enabled computers to outperform humans in a variety of tasks, especially in the areas of computer vision and speech recognition. Yet, AI's path has never been smooth, having essentially fallen apart twice in its lifetime ('winters' of AI), both after periods of popular success ('summers' of AI). We provide a brief rundown of AI's evolution over the course of decades, highlighting its crucial moments and major turning points from inception to the present. In doing so, we attempt to learn, anticipate the future, and discuss what steps may be taken to prevent another 'winter'.
translated by 谷歌翻译